Mirror Descent Meets Fixed Share (and feels no regret)

نویسندگان

  • Nicolò Cesa-Bianchi
  • Pierre Gaillard
  • Gábor Lugosi
  • Gilles Stoltz
چکیده

Mirror descent with an entropic regularizer is known to achieve shifting regretbounds that are logarithmic in the dimension. This is done using either a carefullydesigned projection or by a weight sharing technique. Via a novel unified analysis,we show that these two approaches deliver essentially equivalent bounds on a no-tion of regret generalizing shifting, adaptive, discounted, and other related regrets.Our analysis also captures and extends the generalized weight sharing techniqueof Bousquet and Warmuth, and can be refined in several ways, including improve-ments for small losses and adaptive tuning of parameters.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

A survey of Algorithms and Analysis for Adaptive Online Learning

We present tools for the analysis of Follow-The-Regularized-Leader (FTRL), Dual Averaging, and Mirror Descent algorithms when the regularizer (equivalently, proxfunction or learning rate schedule) is chosen adaptively based on the data. Adaptivity can be used to prove regret bounds that hold on every round, and also allows for data-dependent regret bounds as in AdaGrad-style algorithms (e.g., O...

متن کامل

Analysis Techniques for Adaptive Online Learning

We present tools for the analysis of Follow-The-Regularized-Leader (FTRL), Dual Averaging, and Mirror Descent algorithms when the regularizer (equivalently, proxfunction or learning rate schedule) is chosen adaptively based on the data. Adaptivity can be used to prove regret bounds that hold on every round, and also allows for data-dependent regret bounds as in AdaGrad-style algorithms (e.g., O...

متن کامل

On the Universality of Online Mirror Descent

We show that for a general class of convex online learning problems, Mirror Descent can always achieve a (nearly) optimal regret guarantee.

متن کامل

Sequential Change-Point Detection via Online Convex Optimization

Sequential change-point detection when the distribution parameters are unknown is a fundamental problem in statistics and machine learning. When the post-change parameters are unknown, we consider a set of detection procedures based on sequential likelihood ratios with non-anticipating estimators constructed using online convex optimization algorithms such as online mirror descent, which provid...

متن کامل

On near optimality of one-sample update for joint detection and estimation

Sequential hypothesis test and change-point detection when the distribution parameters 1 are unknown is a fundamental problem in statistics and machine learning. We show that for such 2 problems, detection procedures based on sequential likelihood ratios with simple one-sample update 3 estimates such as online mirror descent are nearly second-order optimal. This means that the upper 4 bound for...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2012